首页> 外文OA文献 >A Factorized Recurrent Neural Network based architecture for medium to large vocabulary Language Modelling
【2h】

A Factorized Recurrent Neural Network based architecture for medium to large vocabulary Language Modelling

机译:基于分解回归神经网络的中到中等体系结构   大词汇量语言建模

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Statistical language models are central to many applications that usesemantics. Recurrent Neural Networks (RNN) are known to produce state of theart results for language modelling, outperforming their traditional n-gramcounterparts in many cases. To generate a probability distribution across avocabulary, these models require a softmax output layer that linearly increasesin size with the size of the vocabulary. Large vocabularies need acommensurately large softmax layer and training them on typical laptops/PCsrequires significant time and machine resources. In this paper we present a newtechnique for implementing RNN based large vocabulary language models thatsubstantially speeds up computation while optimally using the limited memoryresources. Our technique, while building on the notion of factorizing theoutput layer by having multiple output layers, improves on the earlier work bysubstantially optimizing on the individual output layer size and alsoeliminating the need for a multistep prediction process.
机译:统计语言模型对于许多使用语义的应用程序至关重要。众所周知,递归神经网络(RNN)可以为语言建模提供最新的结果,在许多情况下优于传统的n语法对手。为了在整个词汇表中生成概率分布,这些模型需要一个softmax输出层,该输出层的大小随词汇表的大小线性增加。大词汇量需要相应大的softmax层,并且要在典型的笔记本电脑/ PC上对其进行培训需要大量的时间和机器资源。在本文中,我们提出了一种新技术,用于实现基于RNN的大词汇量语言模型,该模型可以显着加快计算速度,同时可以最佳地使用有限的内存资源。我们的技术在通过具有多个输出层来分解输出层的概念的基础上,通过充分优化各个输出层的大小并消除了对多步预测过程的需求,改进了早期的工作。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号